# Sparse Attention
Lsg Legal Small Uncased 4096
A compact version of LEGAL-BERT, employing Local+Sparse+Global attention mechanism (LSG) for efficient long-sequence processing
Large Language Model
Transformers English

L
ccdv
1,088
0
Bigbird Base Trivia Itc
Apache-2.0
A fine-tuned model based on bigbird-roberta-base, optimized for trivia QA tasks with long sequence processing support.
Question Answering System English
B
google
1,049
8
Bigbird Pegasus Large Bigpatent
Apache-2.0
BigBird is a Transformer model based on sparse attention, capable of processing sequences up to 4096 in length, suitable for tasks like long document summarization.
Text Generation
Transformers English

B
google
945
40
Bigbird Roberta Base Finetuned App
MIT
A Transformer model based on sparse attention, fine-tuned specifically for mobile app description classification tasks
Text Classification
Transformers English

B
nsi319
15
0
Bigbird Pegasus Large Arxiv
Apache-2.0
BigBird is a Transformer model based on sparse attention, capable of handling longer sequences, suitable for tasks like long document summarization.
Text Generation
Transformers English

B
google
8,528
61
Bigbird Pegasus Large Pubmed
Apache-2.0
BigBirdPegasus is a Transformer model based on sparse attention, capable of handling longer sequences, especially suitable for long document summarization tasks.
Text Generation
Transformers English

B
google
2,031
47
Bigbird Roberta Large
Apache-2.0
BigBird is a Transformer model based on sparse attention, capable of processing sequences up to 4096 tokens long, suitable for long document tasks.
Large Language Model English
B
google
1,152
27
Featured Recommended AI Models